var runtime.debug
121 uses
runtime (current package)
arena.go#L816: if debug.malloc {
cgocall.go#L431: if debug.dataindependenttiming == 1 && gp.m.isextra {
cgocall.go#L452: if debug.dataindependenttiming == 1 && !ditAlreadySet {
cgocall.go#L538: if !goexperiment.CgoCheck2 && debug.cgocheck == 0 {
cgocall.go#L789: if !goexperiment.CgoCheck2 && debug.cgocheck == 0 {
chan.go#L819: async := debug.asynctimerchan.Load() != 0
chan.go#L837: async := debug.asynctimerchan.Load() != 0
malloc.go#L1028: if debug.malloc {
malloc.go#L1092: if debug.malloc {
malloc.go#L1615: if debug.sbrk != 0 {
mbitmap.go#L1276: if (GOARCH == "amd64" || GOARCH == "arm64") && p == clobberdeadPtr && debug.invalidptr != 0 {
mbitmap.go#L1295: if debug.invalidptr != 0 {
mcleanup.go#L89: if debug.sbrk != 0 {
mem_linux.go#L50: if debug.madvdontneed != 0 && advise != madviseUnsupported {
mem_linux.go#L76: if debug.harddecommit > 0 {
mem_linux.go#L85: if debug.harddecommit > 0 {
mem_linux.go#L179: if debug.disablethp != 0 {
mfinal.go#L451: if debug.sbrk != 0 {
mgc.go#L680: if debug.gcstoptheworld == 1 {
mgc.go#L682: } else if debug.gcstoptheworld == 2 {
mgc.go#L1054: if debug.gccheckmark > 0 {
mgc.go#L1237: if debug.gctrace > 0 {
mgc.go#L1596: if debug.gccheckmark > 0 {
mgc.go#L1619: if debug.gccheckmark > 0 {
mgcmark.go#L1598: if debug.gccheckmark > 0 && span.isFree(objIndex) {
mgcpacer.go#L420: if debug.gcstoptheworld > 0 {
mgcpacer.go#L456: if debug.gcpacertrace > 0 {
mgcpacer.go#L674: if debug.gcpacertrace > 0 {
mgcsweep.go#L175: if debug.gcpacertrace > 0 {
mgcsweep.go#L420: if debug.scavtrace > 0 {
mgcsweep.go#L610: if traceAllocFreeEnabled() || debug.clobberfree != 0 || raceenabled || msanenabled || asanenabled {
mgcsweep.go#L624: if debug.clobberfree != 0 {
mgcsweep.go#L814: if debug.efence > 0 {
mheap.go#L1695: if debug.scavtrace > 0 {
mprof.go#L447: nstk := callers(5, mp.profStack[:debug.profstackdepth])
mprof.go#L534: if debug.profstackdepth == 0 {
mprof.go#L782: if debug.profstackdepth == 0 {
mprof.go#L805: if debug.runtimeContentionStacks.Load() == 0 {
mprof.go#L833: nstk := int(debug.profstackdepth)
panic.go#L744: if debug.panicnil.Load() != 1 {
panic.go#L1373: if debug.schedtrace > 0 || debug.scheddetail > 0 {
preempt.go#L223: if preemptMSupported && debug.asyncpreemptoff == 0 && needAsync {
proc.go#L196: if debug.inittrace != 0 {
proc.go#L350: if debug.gctrace > 0 {
proc.go#L986: if debug.profstackdepth == 0 {
proc.go#L1005: return make([]uintptr, 1+maxSkip+debug.profstackdepth)
proc.go#L1010: func makeProfStack() []uintptr { return make([]uintptr, debug.profstackdepth) }
proc.go#L1075: if debug.dontfreezetheworld > 0 {
proc.go#L1860: if debug.dataindependenttiming == 1 {
proc.go#L4040: if debug.dontfreezetheworld > 0 && freezing.Load() {
proc.go#L5168: if debug.tracebackancestors <= 0 || callergp.goid == 0 {
proc.go#L5176: if n > debug.tracebackancestors {
proc.go#L5177: n = debug.tracebackancestors
proc.go#L6147: if debug.schedtrace <= 0 && (sched.gcwaiting.Load() || sched.npidle.Load() == gomaxprocs) {
proc.go#L6249: if debug.schedtrace > 0 && lasttrace+int64(debug.schedtrace)*1000000 <= now {
proc.go#L6251: schedtrace(debug.scheddetail > 0)
proc.go#L6390: if preemptMSupported && debug.asyncpreemptoff == 0 {
runtime1.go#L309: var debug struct {
runtime1.go#L366: {name: "adaptivestackstart", value: &debug.adaptivestackstart},
runtime1.go#L367: {name: "asyncpreemptoff", value: &debug.asyncpreemptoff},
runtime1.go#L368: {name: "asynctimerchan", atomic: &debug.asynctimerchan},
runtime1.go#L369: {name: "cgocheck", value: &debug.cgocheck},
runtime1.go#L370: {name: "clobberfree", value: &debug.clobberfree},
runtime1.go#L371: {name: "dataindependenttiming", value: &debug.dataindependenttiming},
runtime1.go#L372: {name: "disablethp", value: &debug.disablethp},
runtime1.go#L373: {name: "dontfreezetheworld", value: &debug.dontfreezetheworld},
runtime1.go#L374: {name: "efence", value: &debug.efence},
runtime1.go#L375: {name: "gccheckmark", value: &debug.gccheckmark},
runtime1.go#L376: {name: "gcpacertrace", value: &debug.gcpacertrace},
runtime1.go#L377: {name: "gcshrinkstackoff", value: &debug.gcshrinkstackoff},
runtime1.go#L378: {name: "gcstoptheworld", value: &debug.gcstoptheworld},
runtime1.go#L379: {name: "gctrace", value: &debug.gctrace},
runtime1.go#L380: {name: "harddecommit", value: &debug.harddecommit},
runtime1.go#L381: {name: "inittrace", value: &debug.inittrace},
runtime1.go#L382: {name: "invalidptr", value: &debug.invalidptr},
runtime1.go#L383: {name: "madvdontneed", value: &debug.madvdontneed},
runtime1.go#L384: {name: "panicnil", atomic: &debug.panicnil},
runtime1.go#L385: {name: "profstackdepth", value: &debug.profstackdepth, def: 128},
runtime1.go#L386: {name: "runtimecontentionstacks", atomic: &debug.runtimeContentionStacks},
runtime1.go#L387: {name: "sbrk", value: &debug.sbrk},
runtime1.go#L388: {name: "scavtrace", value: &debug.scavtrace},
runtime1.go#L389: {name: "scheddetail", value: &debug.scheddetail},
runtime1.go#L390: {name: "schedtrace", value: &debug.schedtrace},
runtime1.go#L391: {name: "traceadvanceperiod", value: &debug.traceadvanceperiod},
runtime1.go#L392: {name: "traceallocfree", atomic: &debug.traceallocfree},
runtime1.go#L393: {name: "tracecheckstackownership", value: &debug.traceCheckStackOwnership},
runtime1.go#L394: {name: "tracebackancestors", value: &debug.tracebackancestors},
runtime1.go#L395: {name: "tracefpunwindoff", value: &debug.tracefpunwindoff},
runtime1.go#L400: debug.cgocheck = 1
runtime1.go#L401: debug.invalidptr = 1
runtime1.go#L402: debug.adaptivestackstart = 1 // set this to 0 to turn larger initial goroutine stacks off
runtime1.go#L412: debug.madvdontneed = 1
runtime1.go#L414: debug.traceadvanceperiod = defaultTraceAdvancePeriod
runtime1.go#L440: debug.malloc = (debug.inittrace | debug.sbrk) != 0
runtime1.go#L441: debug.profstackdepth = min(debug.profstackdepth, maxProfStackDepth)
runtime1.go#L530: if debug.cgocheck > 1 {
signal_unix.go#L447: if sig == sigPreempt && preemptMSupported && debug.asyncpreemptoff == 0 {
signal_unix.go#L685: if sig == sigPreempt && debug.asyncpreemptoff == 0 && !delayedSignal {
signal_unix.go#L1391: if sig == sigPreempt && preemptMSupported && debug.asyncpreemptoff == 0 {
stack.go#L351: if debug.efence != 0 || stackFromSystem != 0 {
stack.go#L460: if debug.efence != 0 || stackFromSystem != 0 {
stack.go#L461: if debug.efence != 0 || stackFaultOnFree != 0 {
stack.go#L635: if f.valid() && 0 < p && p < minLegalPointer && debug.invalidptr != 0 {
stack.go#L1206: if debug.gcshrinkstackoff > 0 {
stack.go#L1319: if debug.adaptivestackstart == 0 {
synctest.go#L155: if debug.asynctimerchan.Load() != 0 {
time.go#L473: async := debug.asynctimerchan.Load() != 0
time.go#L554: async := debug.asynctimerchan.Load() != 0
time.go#L1108: async := debug.asynctimerchan.Load() != 0
trace.go#L254: if debug.traceallocfree.Load() != 0 {
trace.go#L256: trace.debugMalloc = debug.malloc
trace.go#L258: debug.malloc = true
trace.go#L628: debug.malloc = trace.debugMalloc
trace.go#L960: s.timer.sleep(int64(debug.traceadvanceperiod))
tracestack.go#L49: if debug.traceCheckStackOwnership != 0 && gp != nil {
tracestack.go#L248: return debug.tracefpunwindoff != 0 || (goarch.ArchFamily != goarch.AMD64 && goarch.ArchFamily != goarch.ARM64)